AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Disentangled Attention

# Disentangled Attention

Erlangshen DeBERTa V2 320M Chinese
Apache-2.0
Chinese pre-trained language model based on DeBERTa-v2 architecture with 320 million parameters, excelling in natural language understanding tasks
Large Language Model Transformers Chinese
E
IDEA-CCNL
186
13
Deberta V3 Base
MIT
DeBERTaV3 is an improved pre-trained language model based on DeBERTa, which enhances efficiency through gradient-disentangled embedding sharing in ELECTRA-style pretraining and excels in natural language understanding tasks.
Large Language Model English
D
microsoft
1.6M
316
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase